# Information retrieval optimization
Minicoil V1
Apache-2.0
MiniCOIL is a sparse contextualized word-by-word embedding model specifically designed for efficient semantic similarity computation
Text Embedding English
M
Qdrant
564
7
Plamo Embedding 1b
Apache-2.0
PLaMo-Embedding-1B is a Japanese text embedding model developed by Preferred Networks, demonstrating outstanding performance in Japanese text embedding benchmarks
Text Embedding
Transformers Japanese

P
pfnet
33.48k
25
Reranker ModernBERT Large Gooaq Bce
Apache-2.0
This is a cross-encoder model fine-tuned from ModernBERT-large, used to calculate the scores of text pairs, suitable for text re-ranking and semantic search tasks.
Text Embedding English
R
tomaarsen
596
5
Reranker Pho BLAI
Apache-2.0
This is a Vietnamese text ranking model based on the Apache-2.0 license, primarily used for Vietnamese text ranking tasks.
Large Language Model Other
R
truong1301
21
0
Hypencoder.2 Layer
Apache-2.0
Hypencoder is a hypernetwork model for information retrieval, consisting of a text encoder and Hypencoder. It can convert text into a small neural network and output relevance scores.
Text Embedding
Transformers English

H
jfkback
18
1
Hypencoder.8 Layer
MIT
Hypencoder is a dual-encoder model for information retrieval, consisting of a text encoder and a hypernetwork (Hypencoder), capable of converting text into small neural networks for computing relevance scores.
Text Embedding
Transformers English

H
jfkback
18
1
Lightblue.lb Reranker 0.5B V1.0 GGUF
A lightweight text ranking model suitable for information retrieval and document sorting tasks.
Large Language Model
L
DevQuasar
66
0
Modernbert Base Msmarco
This model is a fine-tuned sentence embedding model based on ModernBERT-base, specifically designed for sentence similarity tasks and supports English text processing.
Text Embedding English
M
joe32140
4,695
9
Arabic Retrieval V1.0
Apache-2.0
A high-performance Arabic information retrieval model built on the sentence-transformers framework, optimized for the richness and complexity of the Arabic language.
Text Embedding Arabic
A
omarelshehy
366
3
Arabic Reranker
This is an Arabic reranking model based on the BERT architecture, specifically designed for Arabic text, performing reranking tasks by scoring and sorting text options.
Text Embedding Arabic
A
oddadmix
14
0
Polish Reranker Bge V2
This is a reranking model based on BAAI/bge-reranker-v2-m3 and further fine-tuned on a large-scale Polish text pair dataset, supporting long-context processing.
Text Embedding
Transformers Other

P
sdadas
549
1
Phoranker
Apache-2.0
PhoRanker is a cross-encoder model for Vietnamese text ranking, capable of efficiently classifying and ranking Vietnamese texts.
Text Embedding
Transformers Other

P
itdainb
4,063
15
Norwegian Nli Triplets C
Apache-2.0
A Norwegian sentence embedding model fine-tuned based on jina-embeddings-v2-base-en, focusing on keyword document search and sentence similarity tasks
Text Embedding Other
N
fine-tuned
24
1
Crossencoder Me5 Base Mmarcofr
MIT
This is a French cross-encoder model based on multilingual-e5-base, specifically designed for passage reranking tasks.
Text Embedding French
C
antoinelouis
49
1
Crossencoder Camembert Large Mmarcofr
MIT
This is a French cross-encoder model specifically designed for passage re-ranking tasks in semantic search.
Text Embedding French
C
antoinelouis
108
1
Gte Large En V1.5
Apache-2.0
GTE-Large is a high-performance English text embedding model that excels in multiple text similarity and classification tasks.
Text Embedding
Transformers Supports Multiple Languages

G
Alibaba-NLP
891.76k
213
Splade V3 Distilbert
SPLADE-v3-DistilBERT is the DistilBERT version of naver/splade-v3, which performs excellently in tasks such as information retrieval.
Text Embedding
Transformers English

S
naver
1,216
6
Splade V3 Lexical
SPLADE-v3-Lexical is a term-weighted version of the SPLADE model, focusing on information retrieval tasks without expansion at the query end.
Text Embedding
Transformers English

S
naver
1,184
2
Polish Reranker Base Mse
Apache-2.0
This is a Polish text ranking model trained using Mean Squared Error (MSE) distillation method, with a training dataset containing 1.4 million queries and 10 million document text pairs.
Text Embedding
Transformers Other

P
sdadas
16
0
Polish Reranker Large Ranknet
Apache-2.0
This is a Polish text ranking model trained using the RankNet loss function, with a training dataset consisting of 1.4 million queries and 10 million document pairs.
Text Embedding
Transformers Other

P
sdadas
337
2
Minilm L6 Danish Reranker
MIT
This is a lightweight Danish text ranking model adapted from the English MiniLM-L6 model, specifically designed for Danish information retrieval tasks.
Text Embedding Other
M
KennethTM
160
0
Simcse Retromae Small Cs
A small Czech semantic embedding model fine-tuned with SimCSE objective based on RetroMAE-Small
Text Embedding
Transformers Other

S
Seznam
309
4
Mmlw Retrieval Roberta Large
Apache-2.0
MMLW (I Must Get Better Messages) is a neural text encoder for Polish, optimized for information retrieval tasks.
Text Embedding
Transformers Other

M
sdadas
237.90k
12
Mmlw Retrieval E5 Base
Apache-2.0
MMLW (I Must Get Better Messages) is a Polish neural text encoder optimized for information retrieval tasks, capable of converting queries and passages into 768-dimensional vectors.
Text Embedding
Transformers Other

M
sdadas
144
1
Ember V1
MIT
Ember v1 is an embedding model based on sentence-transformers, primarily used for feature extraction and sentence similarity calculation.
Text Embedding
Transformers English

E
llmrails
51.52k
62
Bge Micro
bge_micro is a lightweight sentence similarity calculation model based on transformer architecture, specifically designed for efficient feature extraction and sentence similarity tasks.
Text Embedding
Transformers

B
TaylorAI
1,799
23
Bge Base En V1.5 Ct2
MIT
BGE Base English v1.5 is a transformer-based sentence embedding model, specifically designed for extracting sentence features and calculating sentence similarity.
Text Embedding
Transformers English

B
winstxnhdw
30
0
Crossencoder Camembert Base Mmarcofr
MIT
This is a French cross-encoder model based on CamemBERT, specifically designed for passage reranking tasks, demonstrating excellent performance on the mMARCO-fr dataset.
Text Embedding French
C
antoinelouis
622
5
Biencoder Camembert Base Mmarcofr
MIT
This is a dense single-vector dual-encoder model for French, suitable for semantic search tasks.
Text Embedding French
B
antoinelouis
984
9
Compositional Bert Base Uncased
Apache-2.0
A sentence similarity calculation model based on the CompCSE dedicated dataset, suitable for English text processing.
Text Embedding
Transformers English

C
perceptiveshawty
20
1
Instructor Large
Apache-2.0
INSTRUCTOR is a text embedding model based on the T5 architecture, focusing on sentence similarity calculation and text classification tasks, and supports English language processing.
Text Embedding
Transformers English

I
hkunlp
186.12k
508
Msmarco Distilbert Base Tas B Mmarco Pt 300k
This is a Portuguese sentence embedding model based on the DistilBERT architecture, specifically optimized for semantic similarity tasks.
Text Embedding
Transformers Other

M
mpjan
37
4
Msmarco Distilbert Base Tas B Mmarco Pt 100k
This is a Portuguese sentence transformer model based on DistilBERT, specifically designed for sentence similarity and semantic search tasks.
Text Embedding
Transformers Other

M
mpjan
44
4
Cross Encoder Mmarco German Distilbert Base
Apache-2.0
A German cross-encoder model fine-tuned on the MMARCO dataset for query-passage relevance scoring
Text Embedding German
C
ml6team
1,026
3
Monot5 Large Msmarco 10k
A T5-large reranker fine-tuned for 10,000 steps on the MS MARCO passage dataset, excelling on non-MS MARCO datasets
Large Language Model
M
castorini
168
1
Monot5 Base Msmarco 10k
A reranker based on the T5-base architecture, fine-tuned for 10,000 steps on the MS MARCO passage dataset, with excellent zero-shot performance.
Large Language Model
M
castorini
5,396
14
SGPT 2.7B Weightedmean Msmarco Specb Bitfit
SGPT-2.7B is a sentence transformer model based on the weighted mean method, focusing on sentence similarity tasks, trained on the MSMARCO dataset with BitFit technology applied.
Text Embedding
S
Muennighoff
85
3
Featured Recommended AI Models